AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
DPO Fine-Tuning Optimization

# DPO Fine-Tuning Optimization

Gemma3 27b Abliterated Dpo
A fine-tuned large language model based on mlabonne/gemma-3-27b-it-abliterated, trained using the Unsloth acceleration framework and Huggingface's TRL library, achieving a 2x efficiency improvement.
Large Language Model Transformers English
G
summykai
326
3
Chocolatine Fusion 14B
MIT
Chocolatine-Fusion-14B is a merged model that combines the strengths of the Chocolatine-2 series, enhancing reasoning capabilities and multi-turn dialogue performance through optimized fusion.
Large Language Model Transformers
C
FINGU-AI
226
5
Neuralbeagle14 7B 8.0bpw H8 Exl2
Apache-2.0
NeuralBeagle14-7B is a 7B-parameter large language model fine-tuned using the DPO method based on the Beagle14-7B model, excelling in the 7B parameter category.
Large Language Model Transformers
N
LoneStriker
111
5
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase